471 research outputs found
ALTER: Auxiliary Text Rewriting Tool for Natural Language Generation
In this paper, we describe ALTER, an auxiliary text rewriting tool that
facilitates the rewriting process for natural language generation tasks, such
as paraphrasing, text simplification, fairness-aware text rewriting, and text
style transfer. Our tool is characterized by two features, i) recording of
word-level revision histories and ii) flexible auxiliary edit support and
feedback to annotators. The text rewriting assist and traceable rewriting
history are potentially beneficial to the future research of natural language
generation.Comment: EMNLP 2019 (Demo
DiffGuard: Semantic Mismatch-Guided Out-of-Distribution Detection using Pre-trained Diffusion Models
Given a classifier, the inherent property of semantic Out-of-Distribution
(OOD) samples is that their contents differ from all legal classes in terms of
semantics, namely semantic mismatch. There is a recent work that directly
applies it to OOD detection, which employs a conditional Generative Adversarial
Network (cGAN) to enlarge semantic mismatch in the image space. While achieving
remarkable OOD detection performance on small datasets, it is not applicable to
ImageNet-scale datasets due to the difficulty in training cGANs with both input
images and labels as conditions. As diffusion models are much easier to train
and amenable to various conditions compared to cGANs, in this work, we propose
to directly use pre-trained diffusion models for semantic mismatch-guided OOD
detection, named DiffGuard. Specifically, given an OOD input image and the
predicted label from the classifier, we try to enlarge the semantic difference
between the reconstructed OOD image under these conditions and the original
input image. We also present several test-time techniques to further strengthen
such differences. Experimental results show that DiffGuard is effective on both
Cifar-10 and hard cases of the large-scale ImageNet, and it can be easily
combined with existing OOD detection techniques to achieve state-of-the-art OOD
detection results.Comment: Accepted by ICCV2023, with supplementary material
Bridging Parametric and Nonparametric Methods in Cognitive Diagnosis
A number of parametric and nonparametric methods for estimating cognitive
diagnosis models (CDMs) have been developed and applied in a wide range of
contexts. However, in the literature, a wide chasm exists between these two
families of methods, and their relationship to each other is not well
understood. In this paper, we propose a unified estimation framework to bridge
the divide between parametric and nonparametric methods in cognitive diagnosis
to better understand their relationship. We also develop iterative joint
estimation algorithms and establish consistency properties within the proposed
framework. Lastly, we present comprehensive simulation results to compare
different methods, and provide practical recommendations on the appropriate use
of the proposed framework in various CDM contexts
A Note on Improving Variational Estimation for Multidimensional Item Response Theory
Survey instruments and assessments are frequently used in many domains of
social science. When the constructs that these assessments try to measure
become multifaceted, multidimensional item response theory (MIRT) provides a
unified framework and convenient statistical tool for item analysis,
calibration, and scoring. However, the computational challenge of estimating
MIRT models prohibits its wide use because many of the extant methods can
hardly provide results in a realistic time frame when the number of dimensions,
sample size, and test length are large. Instead, variational estimation
methods, such as Gaussian Variational Expectation Maximization (GVEM)
algorithm, have been recently proposed to solve the estimation challenge by
providing a fast and accurate solution. However, results have shown that
variational estimation methods may produce some bias on discrimination
parameters during confirmatory model estimation, and this note proposes an
importance weighted version of GVEM (i.e., IW-GVEM) to correct for such bias
under MIRT models. We also use the adaptive moment estimation method to update
the learning rate for gradient descent automatically. Our simulations show that
IW-GVEM can effectively correct bias with modest increase of computation time,
compared with GVEM. The proposed method may also shed light on improving the
variational estimation for other psychometrics models
Transferring Cross-domain Knowledge for Video Sign Language Recognition
Word-level sign language recognition (WSLR) is a fundamental task in sign
language interpretation. It requires models to recognize isolated sign words
from videos. However, annotating WSLR data needs expert knowledge, thus
limiting WSLR dataset acquisition. On the contrary, there are abundant
subtitled sign news videos on the internet. Since these videos have no
word-level annotation and exhibit a large domain gap from isolated signs, they
cannot be directly used for training WSLR models. We observe that despite the
existence of a large domain gap, isolated and news signs share the same visual
concepts, such as hand gestures and body movements. Motivated by this
observation, we propose a novel method that learns domain-invariant visual
concepts and fertilizes WSLR models by transferring knowledge of subtitled news
sign to them. To this end, we extract news signs using a base WSLR model, and
then design a classifier jointly trained on news and isolated signs to coarsely
align these two domain features. In order to learn domain-invariant features
within each class and suppress domain-specific features, our method further
resorts to an external memory to store the class centroids of the aligned news
signs. We then design a temporal attention based on the learnt descriptor to
improve recognition performance. Experimental results on standard WSLR datasets
show that our method outperforms previous state-of-the-art methods
significantly. We also demonstrate the effectiveness of our method on
automatically localizing signs from sign news, achieving 28.1 for [email protected]: CVPR2020 (oral) preprin
- …